Locality Optimizations for Parallel Machines
نویسنده
چکیده
This paper focuses on the problem of locality optimizations for high-performance uniprocessor and multiprocessor systems. It shows that the problems of minimizing interprocessor communication and optimizing cache locality can be formulated in a similar manner. It outlines the algorithms to optimize for the various levels of the memory hierarchy simultaneously.
منابع مشابه
Locality Optimizations for Parallel Computing Using Data Access Information
Given the large communication overheads characteristic of modern parallel machines, optimizations that improve locality by executing tasks close to data that they will access may improve the performance of parallel computations. This paper describes our experience automatically applying locality optimizations in the context of Jade, a portable, implicitly parallel programming language designed ...
متن کاملOptimizations for Parallel Computing Using DataAccess
Given the large communication overheads characteristic of modern parallel machines, optimizations that eliminate, hide or parallelize communication may improve the performance of parallel computations. This paper describes our experience automatically applying communication optimizations in the context of Jade, a portable, implicitly parallel programming language designed for exploiting task-le...
متن کاملVolume Driven Data Distribution for NUMA-Machines
Highly scalable parallel computers, e.g. SCI-coupled workstation clusters, are NUMA architectures. Thus good static locality is essential for high performance and scalability of parallel programs on these machines. This paper describes novel techniques to optimize static locality at compilation time by application of data transformations and data distributions. The metric which guides the optim...
متن کاملChapter 1 an Overview of the Suif Compiler for Scalable Parallel Machines
We are building a compiler that automatically translates sequential scientiic programs into parallel code for scalable parallel machines. Many of the compiler techniques needed to generate correct and eecient code are common across all scalable machines, regardless of whether its address space is shared or distributed. This paper describes the structure of the compiler, emphasizing the common a...
متن کاملAn Overview of the SUIF Compiler for Scalable Parallel Machines
We are building a compiler that automatically translates sequential scientiic programs into parallel code for scalable parallel machines. Many of the compiler techniques needed to generate correct and eecient code are common across all scalable machines, regardless of whether its address space is shared or distributed. This paper describes the structure of the compiler, emphasizing the common a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1994